中文 |

Newsroom

Multimodal LLMs Can Develop Human-like Object Concept Representations: Study

Jun 11, 2025

A group of Chinese scientists confirmed that multimodal large language models (LLMs) can spontaneously develop human-like object concept representations, providing a new path for the cognitive science of artificial intelligence (AI) and a theoretical framework for building AI systems with human-like cognitive structures.

With the advent of LLMs such as ChatGPT, scientists have started to wonder whether these models can develop human-like object concept representations from linguistic and multimodal data.

"The ability to conceptualize objects in nature has long been regarded as the core of human intelligence," said He Huiguang, a researcher at the Institute of Automation under the Chinese Academy of Sciences (CAS).

When people see objects like a dog, a car or an apple, they can not only identify their physical features such as size, color and shape, but also understand their functions, emotional values, and cultural significance. This multidimensional concept representation forms the cornerstone of human cognition, added He, also the corresponding author of the study paper published in Nature Machine Intelligence on Monday.

Researchers from institutions including the Institute of Automation and the CAS Center for Excellence in Brain Science and Intelligence Technology, combined behavioural and neuroimaging analyses to explore the relationship between object concept representations in LLMs and human cognition.

They designed an innovative paradigm that integrates computational modeling, behavioral experiments and brain science, and they constructed a conceptual map for LLMs.

The study found that the 66 dimensions extracted from LLMs behavioral data are strongly correlated with the neural activity patterns in human brain's category-selective regions. It also compared the consistency of multiple models with human behavior in terms of choice patterns, with the results showing that multimodal LLMs performed better in terms of consistency. (Xinhua)

Contact

E-mail:

Human-like object concept representations emerge naturally in multimodal large language models

Related Articles
Contact Us
  • 86-10-68597521 (day)

    86-10-68597289 (night)

  • 52 Sanlihe Rd., Xicheng District,

    Beijing, China (100864)

Copyright © 2002 - Chinese Academy of Sciences